Multimodal HALEF: An Open-Source Modular Web-Based Multimodal Dialog Framework
نویسندگان
چکیده
We present an open-source web-based multimodal dialog framework, “Multimodal HALEF”, that integrates video conferencing and telephony abilities into the existing HALEF cloud-based dialog framework via the FreeSWITCH video telephony server. Due to its distributed and cloud-based architecture, Multimodal HALEF allows researchers to collect video and speech data from participants interacting with the dialog system outside of traditional lab settings, therefore largely reducing cost and labor incurred during the traditional audio-visual data collection process. The framework is equipped with a set of tools including a web-based user survey template, a speech transcription, annotation and rating portal, a web visual processing server that performs head tracking, and a database that logs full-call audio and video recordings as well as other call-specific information. We present observations from an initial data collection based on an job interview application. Finally we report on some future plans for development of the framework.
منابع مشابه
A distributed cloud-based dialog system for conversational application development
We have previously presented HALEF– an open-source spoken dialog system–that supports telephonic interfaces and has a distributed architecture. In this paper, we extend this infrastructure to be cloudbased, and thus truly distributed and scalable. This cloud-based spoken dialog system can be accessed both via telephone interfaces as well as through web clients with WebRTC/HTML5 integration, all...
متن کاملHALEF: An Open-Source Standard-Compliant Telephony-Based Modular Spoken Dialog System: A Review and An Outlook
We describe completed and ongoing research on HALEF, a telephonybased open-source spoken dialog system that can be used with different plug-andplay back-end modules. We present two examples of such a module, one which classifies whether the person calling into the system is intoxicated or not, and the other a question answering application. The system is compliant with World Wide Web Consortium...
متن کاملAn integration framework for a mobile multimodal dialogue system accessing the semantic web
Advanced intelligent multimodal interface systems usually comprise many sub-systems. For the integration of already existing software components in the SMARTWEB system we developed an integration framework, the IHUB. It allows us to reuse already existing components for interpretation and processing of multimodal user interactions. The framework facilitates the integration of the user in the in...
متن کاملConnecting Foundational Ontologies with MPEG-7 Ontologies for Multimodal QA
In the SMARTWEB project [1] we aim at developing a context-aware, mobile, and multimodal interface to the Semantic Web. In order to reach this goal we provide a integrated ontological framework offering coverage for deep semantic content, including ontological representation of multimedia based on the MPEG-7 standard. A discourse ontology covers concepts for multimodal interaction by means of a...
متن کاملCrowdsourcing Multimodal Dialog Interactions: Lessons Learned from the HALEF Case
We present a retrospective on collecting data of human interactions with multimodal dialog systems (“dialog data”) using crowdsourcing techniques. This is largely based on our experience using the HALEF multimodal dialog system to deploy education-domain conversational applications on the Amazon Mechanical Turk crowdsourcing platform. We list the various lessons learned from this endeavor over ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016